-
Notifications
You must be signed in to change notification settings - Fork 697
Add option to disable operator profiling #5720
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/5720
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 0062774 with merge base 3a25651 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
9602572 to
73ee018
Compare
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: Vysarat Differential Revision: D61883224
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Differential Revision: D61883224
73ee018 to
114b46d
Compare
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Differential Revision: D61883224
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Differential Revision: D61883224
114b46d to
fd47942
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Differential Revision: D61883224
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Differential Revision: D61883224
fd47942 to
14fb389
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Differential Revision: D61883224
14fb389 to
6467980
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
6467980 to
ce1891b
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
e24ae6f to
9fdafe1
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
9fdafe1 to
e520deb
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
e520deb to
84f0087
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Reviewed By: dbort Differential Revision: D61883224
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
84f0087 to
db51eca
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
db51eca to
93b98d2
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Reviewed By: dbort Differential Revision: D61883224
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Reviewed By: dbort Differential Revision: D61883224
Summary: X-link: pytorch/pytorch#136838 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Reviewed By: dbort Differential Revision: D61883224
93b98d2 to
0062774
Compare
|
This pull request was exported from Phabricator. Differential Revision: D61883224 |
|
This pull request has been merged in acfcdd5. |
Summary: X-link: pytorch/executorch#5720 For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time. To disable operator profiling users need to do: ``` etdump_gen.set_event_tracer_profiling_level(executorch::runtime::EventTracerProfilingLevel::kNoOperatorProfiling); ``` Test Plan: Added test case. Differential Revision: D61883224 Pull Request resolved: #136838 Approved by: https://github.com/dbort
Summary:
For smaller models the overhead of profiling ops might be prohibitively large (distorting the inference execution time significantly) so we provide users an option to disable op profiling and essentially only profile the important events such as inference execution time.
To disable operator profiling users need to do:
Differential Revision: D61883224